Bagging Decision Multi-trees

نویسندگان

  • Vicent Estruch
  • César Ferri
  • José Hernández-Orallo
  • M. José Ramírez-Quintana
چکیده

Ensemble methods improve accuracy by combining the predictions of a set of different hypotheses. A well-known method for generating hypothesis ensembles is Bagging. One of the main drawbacks of ensemble methods in general, and Bagging in particular, is the huge amount of computational resources required to learn, store, and apply the set of models. Another problem is that even using the bootstrap technique, many simple models are similar, so limiting the ensemble diversity. In this work, we investigate an optimization technique based on sharing the common parts of the models from an ensemble formed by decision trees in order to minimize both problems. Concretely, we employ a structure called decision multi-tree which can contain simultaneously a set of decision trees and hence consider just once the “repeated” parts. A thorough experimental evaluation is included to show that the proposed optimisation technique pays off in practice.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Binary and Multi-class Classification of fused LIDAR-Imagery Data using an Ensemble Method

Airborne Light Detection and Ranging (LIDAR) data is used for multiple applications, such as urban planning, emergency response, flood control, and city 3D reconstruction. The LIDAR data in its raw form needs to be classified for the above applications. There are two types of classifications: binary and multi-class. In the binary classification, the given LIDAR data is classified into two class...

متن کامل

A Bagging Method using Decision Trees in the Role of Base Classifiers

This paper describes a set of experiments with bagging – a method, which can improve results of classification algorithms. Our use of this method aims at classification algorithms generating decision trees. Results of performance tests focused on the use of the bagging method on binary decision trees are presented. The minimum number of decision trees, which enables an improvement of the classi...

متن کامل

Parallelizing Boosting and Bagging

Bagging and boosting are two general techniques for building predictors based on small samples from a dataset. We show that boosting can be parallelized, and then present performance results for parallelized bagging and boosting using OC1 decision trees and two standard datasets. The main results are that sample sizes limit achievable accuracy, regardless of computational time spent; that paral...

متن کامل

Machine Learning Ensembles: An Empirical Study and Novel Approach

Two learning ensemble methods, Bagging and Boosting, have been applied to decision trees to improve classification accuracy over that of a single decision tree learner. We introduce Bagging and propose a variant of it — Improved Bagging — which, in general, outperforms the original bagging algorithm. We experiment on 22 datasets from the UCI repository, with emphasis on the ensemble’s accuracy ...

متن کامل

{19 () Improving Bagging Performance by Increasing Decision Tree Diversity

Ensembles of decision trees often exhibit greater predictive accuracy than single trees alone. Bagging and boosting are two standard ways of generating and combining multiple trees. Boosting has been empirically determined to be the more eeective of the two, and it has recently been proposed that this may be because it produces more diverse trees than bagging. This paper reports empirical nding...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2004